indian context
BharatBBQ: A Multilingual Bias Benchmark for Question Answering in the Indian Context
Tomar, Aditya, Sahoo, Nihar Ranjan, Bhattacharyya, Pushpak
Evaluating social biases in language models (LMs) is crucial for ensuring fairness and minimizing the reinforcement of harmful stereotypes in AI systems. Existing benchmarks, such as the Bias Benchmark for Question Answering (BBQ), primarily focus on Western contexts, limiting their applicability to the Indian context. To address this gap, we introduce BharatBBQ, a culturally adapted benchmark designed to assess biases in Hindi, English, Marathi, Bengali, Tamil, Telugu, Odia, and Assamese. BharatBBQ covers 13 social categories, including 3 intersectional groups, reflecting prevalent biases in the Indian sociocultural landscape. Our dataset contains 49,108 examples in one language that are expanded using translation and verification to 392,864 examples in eight different languages. We evaluate five multilingual LM families across zero and few-shot settings, analyzing their bias and stereotypical bias scores. Our findings highlight persistent biases across languages and social categories and often amplified biases in Indian languages compared to English, demonstrating the necessity of linguistically and culturally grounded benchmarks for bias evaluation.
- North America > Mexico > Mexico City > Mexico City (0.04)
- North America > United States > Washington > King County > Seattle (0.04)
- North America > Canada > Ontario > Toronto (0.04)
- (11 more...)
ASCenD-BDS: Adaptable, Stochastic and Context-aware framework for Detection of Bias, Discrimination and Stereotyping
Bahl, Rajiv, N, Venkatesan, Aglawe, Parimal, Sarasapalli, Aastha, Kancharla, Bhavya, kolukuluri, Chaitanya, Mohite, Harish, Hora, Japneet, Kakollu, Kiran, Diman, Rahul, Kapale, Shubham, Kathula, Sri Bhagya, Motru, Vamsikrishna, Reddy, Yogeshwar
The rapid evolution of Large Language Models (LLMs) has transformed natural language processing but raises critical concerns about biases inherent in their deployment and use across diverse linguistic and sociocultural contexts. This paper presents a framework named ASCenD BDS (Adaptable, Stochastic and Context-aware framework for Detection of Bias, Discrimination and Stereotyping). The framework presents approach to detecting bias, discrimination, stereotyping across various categories such as gender, caste, age, disability, socioeconomic status, linguistic variations, etc., using an approach which is Adaptive, Stochastic and Context-Aware. The existing frameworks rely heavily on usage of datasets to generate scenarios for detection of Bias, Discrimination and Stereotyping. Examples include datasets such as Civil Comments, Wino Gender, WinoBias, BOLD, CrowS Pairs and BBQ. However, such an approach provides point solutions. As a result, these datasets provide a finite number of scenarios for assessment. The current framework overcomes this limitation by having features which enable Adaptability, Stochasticity, Context Awareness. Context awareness can be customized for any nation or culture or sub-culture (for example an organization's unique culture). In this paper, context awareness in the Indian context has been established. Content has been leveraged from Indian Census 2011 to have a commonality of categorization. A framework has been developed using Category, Sub-Category, STEM, X-Factor, Synonym to enable the features for Adaptability, Stochasticity and Context awareness. The framework has been described in detail in Section 3. Overall 800 plus STEMs, 10 Categories, 31 unique SubCategories were developed by a team of consultants at Saint Fox Consultancy Private Ltd. The concept has been tested out in SFCLabs as part of product development.
- Asia > India (0.08)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- North America > United States > Washington > King County > Seattle (0.04)
- Asia > Singapore (0.04)
L3Cube-IndicQuest: A Benchmark Questing Answering Dataset for Evaluating Knowledge of LLMs in Indic Context
Rohera, Pritika, Ginimav, Chaitrali, Salunke, Akanksha, Sawant, Gayatri, Joshi, Raviraj
Large Language Models (LLMs) have made significant progress in incorporating Indic languages within multilingual models. However, it is crucial to quantitatively assess whether these languages perform comparably to globally dominant ones, such as English. Currently, there is a lack of benchmark datasets specifically designed to evaluate the regional knowledge of LLMs in various Indic languages. In this paper, we present the L3Cube-IndicQuest, a gold-standard question-answering benchmark dataset designed to evaluate how well multilingual LLMs capture regional knowledge across various Indic languages. The dataset contains 200 question-answer pairs, each for English and 19 Indic languages, covering five domains specific to the Indic region. We aim for this dataset to serve as a benchmark, providing ground truth for evaluating the performance of LLMs in understanding and representing knowledge relevant to the Indian context. The IndicQuest can be used for both reference-based evaluation and LLM-as-a-judge evaluation. The dataset is shared publicly at https://github.com/l3cube-pune/indic-nlp .
DataLike: Interview with Sarah Masud
Sarah Masud is a fifth-year PhD scholar at the Laboratory for Computational Social Systems (LCS2) at the Indraprastha Institute of Information Technology, Delhi (IIIT-D). She holds the prestigious Google PhD Fellowship (2023-present) was previously awarded the Prime Minister's Doctoral Fellowship (2020-2023). As part of her PhD, she has authored publications in top-tier venues, addressing the analysis of hateful content in online forums. AI Membership Committee and is a Journal of Open Source Software reviewer. Before her academic pursuits, Sarah worked as a data scientist in developer tooling at Red Hat, Bangalore, for 2.5 years.
- Information Technology > Software (0.56)
- Information Technology > Artificial Intelligence > Natural Language (0.31)
IndiBias: A Benchmark Dataset to Measure Social Biases in Language Models for Indian Context
Sahoo, Nihar Ranjan, Kulkarni, Pranamya Prashant, Asad, Narjis, Ahmad, Arif, Goyal, Tanu, Garimella, Aparna, Bhattacharyya, Pushpak
The pervasive influence of social biases in language data has sparked the need for benchmark datasets that capture and evaluate these biases in Large Language Models (LLMs). Existing efforts predominantly focus on English language and the Western context, leaving a void for a reliable dataset that encapsulates India's unique socio-cultural nuances. To bridge this gap, we introduce IndiBias, a comprehensive benchmarking dataset designed specifically for evaluating social biases in the Indian context. We filter and translate the existing CrowS-Pairs dataset to create a benchmark dataset suited to the Indian context in Hindi language. Additionally, we leverage LLMs including ChatGPT and InstructGPT to augment our dataset with diverse societal biases and stereotypes prevalent in India. The included bias dimensions encompass gender, religion, caste, age, region, physical appearance, and occupation. We also build a resource to address intersectional biases along three intersectional dimensions. Our dataset contains 800 sentence pairs and 300 tuples for bias measurement across different demographics. The dataset is available in English and Hindi, providing a size comparable to existing benchmark datasets. Furthermore, using IndiBias we compare ten different language models on multiple bias measurement metrics. We observed that the language models exhibit more bias across a majority of the intersectional groups.
- North America > United States > New York > New York County > New York City (0.04)
- Europe > Ireland > Leinster > County Dublin > Dublin (0.04)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.04)
- (6 more...)
Casteist but Not Racist? Quantifying Disparities in Large Language Model Bias between India and the West
Khandelwal, Khyati, Tonneau, Manuel, Bean, Andrew M., Kirk, Hannah Rose, Hale, Scott A.
Large Language Models (LLMs), now used daily by millions of users, can encode societal biases, exposing their users to representational harms. A large body of scholarship on LLM bias exists but it predominantly adopts a Western-centric frame and attends comparatively less to bias levels and potential harms in the Global South. In this paper, we quantify stereotypical bias in popular LLMs according to an Indian-centric frame and compare bias levels between the Indian and Western contexts. To do this, we develop a novel dataset which we call Indian-BhED (Indian Bias Evaluation Dataset), containing stereotypical and anti-stereotypical examples for caste and religion contexts. We find that the majority of LLMs tested are strongly biased towards stereotypes in the Indian context, especially as compared to the Western context. We finally investigate Instruction Prompting as a simple intervention to mitigate such bias and find that it significantly reduces both stereotypical and anti-stereotypical biases in the majority of cases for GPT-3.5. The findings of this work highlight the need for including more diverse voices when evaluating LLMs.
- North America > United States > Washington > King County > Seattle (0.14)
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- North America > United States > New York > New York County > New York City (0.04)
- (6 more...)
- Overview (1.00)
- Research Report > Experimental Study (0.46)
Cultural Re-contextualization of Fairness Research in Language Technologies in India
Bhatt, Shaily, Dev, Sunipa, Talukdar, Partha, Dave, Shachi, Prabhakaran, Vinodkumar
Recent research has revealed undesirable biases in NLP data and models. However, these efforts largely focus on social disparities in the West, and are not directly portable to other geo-cultural contexts. In this position paper, we outline a holistic research agenda to re-contextualize NLP fairness research for the Indian context, accounting for Indian societal context, bridging technological gaps in capability and resources, and adapting to Indian cultural values. We also summarize findings from an empirical study on various social biases along different axes of disparities relevant to India, demonstrating their prevalence in corpora and models.
- North America > United States > New York > New York County > New York City (0.04)
- North America > Canada (0.04)
- Europe > France > Provence-Alpes-Côte d'Azur > Bouches-du-Rhône > Marseille (0.04)
- (5 more...)
Re-contextualizing Fairness in NLP: The Case of India
Bhatt, Shaily, Dev, Sunipa, Talukdar, Partha, Dave, Shachi, Prabhakaran, Vinodkumar
Recent research has revealed undesirable biases in NLP data and models. However, these efforts focus on social disparities in West, and are not directly portable to other geo-cultural contexts. In this paper, we focus on NLP fair-ness in the context of India. We start with a brief account of the prominent axes of social disparities in India. We build resources for fairness evaluation in the Indian context and use them to demonstrate prediction biases along some of the axes. We then delve deeper into social stereotypes for Region andReligion, demonstrating its prevalence in corpora and models. Finally, we outline a holistic research agenda to re-contextualize NLP fairness research for the Indian context, ac-counting for Indian societal context, bridging technological gaps in NLP capabilities and re-sources, and adapting to Indian cultural values. While we focus on India, this framework can be generalized to other geo-cultural contexts.
- North America > United States > New York > New York County > New York City (0.04)
- Asia > India > Rajasthan (0.04)
- Asia > India > Mizoram (0.04)
- (12 more...)
- Health & Medicine (0.93)
- Government (0.68)
Delving deep into potential of Artificial Intelligence
Artificial Intelligence and the Future of Power: 5 Battlegrounds" is a game-changing new book by Rajiv Malhotra, analyzing society's vulnerabilities to the current AI revolution. In an Indian context, this is a one of a kind book, covering critical issues which most leaders and policy makers find uncomfortable, and would prefer to sweep under the rug. The author starts out by explaining how he scopes Artificial Intelligence. The topic of AI evokes extreme reactions from people some view it as a technological revolution that will usher an era of unprecedented progress and prosperity. Others tend to equate AI with the takeover of the world and of human kind by evil robots.
- Government (0.44)
- Health & Medicine (0.32)
The Challenges of managing the post-millennials
Going by the pains and frustrations of economic headwinds, most people tend to dismiss it as another wasted year. In this milieu, we tend to overlook the tremendous significance of 2018-19 as an inflection point in workforce composition. Those born at the dawn of the new millennium have turned 18 and are at the threshold of the job market. Their share in the workforce is bound to increase over the next decade. The ability of the cohorts to absorb technology changes and the associated cultural implications is getting sharper with each succeeding generation.